CNNFlow: Memory-driven Data Flow Optimization for Convolutional Neural Networks

نویسندگان

چکیده

Convolution Neural Networks (CNNs) are widely deployed in computer vision applications. The datasets large, and the data reuse across different parts is heavily interleaved. Given that memory access (SRAM especially DRAM) more expensive both performance energy than computation, maximizing to reduce movement hierarchy critical improving execution efficiency. This even important for common use case of CNNs on mobile devices where computing/memory resources limited. We propose CNNFlow, a memory-driven dataflow optimization framework automatically schedule CNN computation given architecture maximize at each level hierarchy. provide mathematical calculation reuses terms parameters including loop ordering, blocking, memory-bank allocation tensors CNN. then present series techniques help prune large search space cost exploration. provides, first time, an exact practical algorithm optimal solutions minimize efficacy demonstrated two used algorithms: AlexNet VGG16 with 5 13 convolution layers, respectively. CNNFlow finds solution layer within tens minutes compute time. Its requires about 20% fewer DRAM accesses 40%–80% SRAM compared state-of-the-art algorithms literature.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Distributed asynchronous optimization of convolutional neural networks

Recently, deep Convolutional Neural Networks have been shown to outperform Deep Neural Networks for acoustic modelling, producing state-of-the-art accuracy in speech recognition tasks. Convolutional models provide increased model robustness through the usage of pooling invariance and weight sharing across spectrum and time. However, training convolutional models is a very computationally expens...

متن کامل

Neuroannealing Martingale-driven Optimization for Neural Networks

Neural networks are effective tools to solve prediction, modeling, and control tasks. However, methods to train neural networks have been less successful on control problems that require the network to model intricately structured regions in state space. This paper presents neuroannealing, a method for training neural network controllers on such problems. Neuroannealing is based on evolutionary...

متن کامل

Model-driven Simulations for Deep Convolutional Neural Networks

The use of simulated virtual environments to train deep convolutional neural networks (CNN) is a currently active practice to reduce the (real)data-hungriness of the deep CNN models, especially in application domains in which large scale real data and/or groundtruth acquisition is difficult or laborious. Recent approaches have attempted to harness the capabilities of existing video games, anima...

متن کامل

Convolutional Neural Networks using Logarithmic Data Representation

Recent advances in convolutional neural networks have considered model complexity and hardware efficiency to enable deployment onto embedded systems and mobile devices. For example, it is now well-known that the arithmetic operations of deep networks can be encoded down to 8-bit fixed-point without significant deterioration in performance. However, further reduction in precision down to as low ...

متن کامل

Central focused convolutional neural networks: Developing a data-driven model for lung nodule segmentation

Accurate lung nodule segmentation from computed tomography (CT) images is of great importance for image-driven lung cancer analysis. However, the heterogeneity of lung nodules and the presence of similar visual characteristics between nodules and their surroundings make it difficult for robust nodule segmentation. In this study, we propose a data-driven model, termed the Central Focused Convolu...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: ACM Transactions on Design Automation of Electronic Systems

سال: 2023

ISSN: ['1084-4309', '1557-7309']

DOI: https://doi.org/10.1145/3577017